Maximizing Diversity Among k-NN Classifiers, An Experimental Study

نویسنده

  • Fuad M. Alkoot
چکیده

Bagging and random feature subsets are used simultaneously to achieve maximum diversity among k-NN component experts in a fusion system. Experimental results indicate that for certain data the combination of the two design methods in a fusion system is beneficial. We also compare two random feature subset design methods, a widely used, conventional, expert based method and a system based design method proposed by us. The system based method shows superior performance when more than 20 percent of the features have discriminatory information..

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Disturbing Neighbors Diversity for Decision Forests

Ensemble methods take their output from a set of base predictors. The ensemble accuracy depends on two factors: the base classifiers accuracy and their diversity (how different are these base classifiers outputs from each other). An approach for increasing the diversity of the base classifiers is presented in this paper. The method builds some new features to be added to the base classifier tra...

متن کامل

Bayesian Network Classifiers Versus k-NN Classifier Using Sequential Feature Selection

The aim of this paper is to compare Bayesian network classifiers to the k-NN classifier based on a subset of features. This subset is established by means of sequential feature selection methods. Experimental results show that Bayesian network classifiers more often achieve a better classification rate on different data sets than selective k-NN classifiers. The k-NN classifier performs well in ...

متن کامل

A New Fuzzy Stacked Generalization Technique and Analysis of its Performance

In this study, a new Stacked Generalization technique called Fuzzy Stacked Generalization (FSG) is proposed to minimize the difference between N -sample and large-sample classification error of the Nearest Neighbor classifier. The proposed FSG employs a new hierarchical distance learning strategy to minimize the error difference. For this purpose, we first construct an ensemble of base-layer fu...

متن کامل

Time-Series Classification in Many Intrinsic Dimensions

In the context of many data mining tasks, high dimensionality was shown to be able to pose significant problems, commonly referred to as different aspects of the curse of dimensionality. In this paper, we investigate in the time-series domain one aspect of the dimensionality curse called hubness, which refers to the tendency of some instances in a data set to become hubs by being included in un...

متن کامل

An Experimental Evaluation of Nearest Neighbour Time Series Classification

Data mining research into time series classification (TSC) has focussed on alternative distance measures for nearest neighbour classifiers. It is standard practice to use 1-NN with Euclidean or dynamic time warping (DTW) distance as a straw man for comparison. As part of a wider investigation into elastic distance measures for TSC [1], we perform a series of experiments to test whether this sta...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2002